Goto

Collaborating Authors

 axiom 2




Differential privacy from axioms

Blanc, Guy, Pires, William, Pitassi, Toniann

arXiv.org Artificial Intelligence

Differential privacy (DP) is the de facto notion of privacy both in theory and in practice. However, despite its popularity, DP imposes strict requirements which guard against strong worst-case scenarios. For example, it guards against seemingly unrealistic scenarios where an attacker has full information about all but one point in the data set, and still nothing can be learned about the remaining point. While preventing such a strong attack is desirable, many works have explored whether average-case relaxations of DP are easier to satisfy [HWR13,WLF16,BF16,LWX23]. In this work, we are motivated by the question of whether alternate, weaker notions of privacy are possible: can a weakened privacy notion still guarantee some basic level of privacy, and on the other hand, achieve privacy more efficiently and/or for a substantially broader set of tasks? Our main result shows the answer is no: even in the statistical setting, any reasonable measure of privacy satisfying nontrivial composition is equivalent to DP. To prove this, we identify a core set of four axioms or desiderata: pre-processing invariance, prohibition of blatant non-privacy, strong composition, and linear scalability. Our main theorem shows that any privacy measure satisfying our axioms is equivalent to DP, up to polynomial factors in sample complexity. We complement this result by showing our axioms are minimal: removing any one of our axioms enables ill-behaved measures of privacy.



llmSHAP: A Principled Approach to LLM Explainability

Naudot, Filip, Sundqvist, Tobias, Kampik, Timotheus

arXiv.org Artificial Intelligence

The rise of data-driven algorithms and, most notably, applications of deep learning has led to concerns about a lack of thorough human oversight of socially important decisions that are either delegated in their entirety to machines, or made by humans based on machine recommendations. Explainable AI (XAI) approaches attempt to mitigate these concerns by helping (typically human) users understand how and why algorithms produce specific outputs [1]. An important class of XAI methods focuses on providing explanations of black-box classifiers that attribute classification outcomes (which one may consider decisions or decision recommendations) to input characteristics (feature values) [2, 3]. Such feature attribution methods can be considered meta-reasoning functions that approximate classifier behavior with the objective of providing users a reasonably faithful intuition of behavioral fundamentals. One of the most prominent feature attribution methods is SHAP, which is based on the Shapley value in cooperative game theory that quantifies players' (feature values') contributions to coalition utility (classification outcomes) [4]. Feature attribution methods have, in general, limitations: notably, they are necessarily approximations, and as purely technical tools, they cannot fully consider crucial nuances of the socio-technical systems they are embedded in [5]; for example, the visualizations provided out-of-the-box by feature attribution software libraries may be difficult to interpret [6]. Still, Shapley value-based approaches can be considered a reasonable choice for facilitating black-box explainability, notably because (i) they are based on well-established and intuitive mathematical principles of the Shapley value and (ii) there is at least some evidence of their potential usefulness, also relative to alternative approaches [6]. However, the Shapley value cannot straight-forwardly be applied to inference from Large Language Models (LLMs), which power many of the currently emerging AI applications.



Clone-Resistant Weights in Metric Spaces: A Framework for Handling Redundancy Bias

Berriaud, Damien, Wattenhofer, Roger

arXiv.org Artificial Intelligence

We are given a set of elements in a metric space. The distribution of the elements is arbitrary, possibly adversarial. Can we weigh the elements in a way that is resistant to such (adversarial) manipulations? This problem arises in various contexts. For instance, the elements could represent data points, requiring robust domain adaptation. Alternatively, they might represent tasks to be aggregated into a benchmark; or questions about personal political opinions in voting advice applications. This article introduces a theoretical framework for dealing with such problems. We propose clone-proof representation functions as a solution concept. These functions distribute importance across elements of a set such that similar objects (``clones'') share (some of) their weights, thus avoiding a potential bias introduced by their multiplicity. Our framework extends the maximum uncertainty principle to accommodate general metric spaces and includes a set of axioms - symmetry, continuity, and clone-proofness - that guide the construction of representation functions. Finally, we address the existence of representation functions satisfying our axioms in the significant case of Euclidean spaces and propose a general method for their construction.


Deontic Temporal Logic for Formal Verification of AI Ethics

V., Priya T., Rao, Shrisha

arXiv.org Artificial Intelligence

Ensuring ethical behavior in Artificial Intelligence (AI) systems amidst their increasing ubiquity and influence is a major concern the world over. The use of formal methods in AI ethics is a possible crucial approach for specifying and verifying the ethical behavior of AI systems. This paper proposes a formalization based on deontic logic to define and evaluate the ethical behavior of AI systems, focusing on system-level specifications, contributing to this important goal. It introduces axioms and theorems to capture ethical requirements related to fairness and explainability. The formalization incorporates temporal operators to reason about the ethical behavior of AI systems over time. The authors evaluate the effectiveness of this formalization by assessing the ethics of the real-world COMPAS and loan prediction AI systems. Various ethical properties of the COMPAS and loan prediction systems are encoded using deontic logical formulas, allowing the use of an automated theorem prover to verify whether these systems satisfy the defined properties. The formal verification reveals that both systems fail to fulfill certain key ethical properties related to fairness and non-discrimination, demonstrating the effectiveness of the proposed formalization in identifying potential ethical issues in real-world AI applications.


Foundations of the Theory of Performance-Based Ranking

Piérard, Sébastien, Halin, Anaïs, Cioppa, Anthony, Deliège, Adrien, Van Droogenbroeck, Marc

arXiv.org Artificial Intelligence

Ranking entities such as algorithms, devices, methods, or models based on their performances, while accounting for application-specific preferences, is a challenge. To address this challenge, we establish the foundations of a universal theory for performance-based ranking. First, we introduce a rigorous framework built on top of both the probability and order theories. Our new framework encompasses the elements necessary to (1) manipulate performances as mathematical objects, (2) express which performances are worse than or equivalent to others, (3) model tasks through a variable called satisfaction, (4) consider properties of the evaluation, (5) define scores, and (6) specify application-specific preferences through a variable called importance. On top of this framework, we propose the first axiomatic definition of performance orderings and performance-based rankings. Then, we introduce a universal parametric family of scores, called ranking scores, that can be used to establish rankings satisfying our axioms, while considering application-specific preferences. Finally, we show, in the case of two-class classification, that the family of ranking scores encompasses well-known performance scores, including the accuracy, the true positive rate (recall, sensitivity), the true negative rate (specificity), the positive predictive value (precision), and F1. However, we also show that some other scores commonly used to compare classifiers are unsuitable to derive performance orderings satisfying the axioms. Therefore, this paper provides the computer vision and machine learning communities with a rigorous framework for evaluating and ranking entities.


Reviews: Clustering Redemption–Beyond the Impossibility of Kleinberg's Axioms

Neural Information Processing Systems

This paper extends Kleinberg's work on the impossibility of clustering. That is, Kleinberg introduced 3 axioms that any clustering procedure should satisfy and then showed that it is impossible to satisfy all three simultaneously. This work suggests a refinement of Kleinberg's 3rd axiom having to do with consistency. In the original axiom, if the clustering distance function is perturbed such that all within-cluster distances do not increase and all across cluster distances do not decrease, then the optimal clustering should be the same with respect to both the perturbed and unperturbed distance functions. The refined consistency axiom proposed in this work says that under the perturbed distance function, the optimal clustering may be different so long as the number of clusters in the optimal partition is different as well.